natural language processing model
Inducing brain-relevant bias in natural language processing models
Progress in natural language processing (NLP) models that estimate representations of word sequences has recently been leveraged to improve the understanding of language processing in the brain. However, these models have not been specifically designed to capture the way the brain represents language meaning. We hypothesize that fine-tuning these models to predict recordings of brain activity of people reading text will lead to representations that encode more brain-activity-relevant language information. We demonstrate that a version of BERT, a recently introduced and powerful language model, can improve the prediction of brain activity after fine-tuning. We show that the relationship between language and brain activity learned by BERT during this fine-tuning transfers across multiple participants. We also show that, for some participants, the fine-tuned representations learned from both magnetoencephalography (MEG) and functional magnetic resonance imaging (fMRI) are better for predicting fMRI than the representations learned from fMRI alone, indicating that the learned representations capture brain-activity-relevant information that is not simply an artifact of the modality. While changes to language representations help the model predict brain activity, they also do not harm the model's ability to perform downstream NLP tasks. Our findings are notable for research on language understanding in the brain.
- Health & Medicine > Health Care Technology (1.00)
- Health & Medicine > Diagnostic Medicine > Imaging (0.60)
Reviews: Inducing brain-relevant bias in natural language processing models
The reviewers agree that this paper makes an original contribution in investigating the fine tuning of a contextual embedding model, e.g. They have identified some issues with the clarity of the motivation of the work and the presentation of some of the results, but I feel that these shortcomings are outweighed by the contributions of this work which should be of interest to researchers across a number of disciplines. The reviewers thank the authors for the additional analysis provided in their response and look forward to this being incorporated into the final paper. In particular it would be good to clarify to significance of the results in Table 1, whether any of the differences are interpretable or whether this simply shows that they are equivalent.
Robust Speech and Natural Language Processing Models for Depression Screening
Lu, Y., Harati, A., Rutowski, T., Oliveira, R., Chlebek, P., Shriberg, E.
Depression is a global health concern with a critical need for increased patient screening. Speech technology offers advantages for remote screening but must perform robustly across patients. We have described two deep learning models developed for this purpose. One model is based on acoustics; the other is based on natural language processing. Both models employ transfer learning. Data from a depression-labeled corpus in which 11,000 unique users interacted with a human-machine application using conversational speech is used. Results on binary depression classification have shown that both models perform at or above AUC=0.80 on unseen data with no speaker overlap. Performance is further analyzed as a function of test subset characteristics, finding that the models are generally robust over speaker and session variables. We conclude that models based on these approaches offer promise for generalized automated depression screening.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- North America > United States > Virginia > Arlington County > Arlington (0.04)
- North America > United States > Texas (0.04)
- (2 more...)
- Health & Medicine > Therapeutic Area > Psychiatry/Psychology (1.00)
- Health & Medicine > Consumer Health (0.88)
Inducing brain-relevant bias in natural language processing models
Progress in natural language processing (NLP) models that estimate representations of word sequences has recently been leveraged to improve the understanding of language processing in the brain. However, these models have not been specifically designed to capture the way the brain represents language meaning. We hypothesize that fine-tuning these models to predict recordings of brain activity of people reading text will lead to representations that encode more brain-activity-relevant language information. We demonstrate that a version of BERT, a recently introduced and powerful language model, can improve the prediction of brain activity after fine-tuning. We show that the relationship between language and brain activity learned by BERT during this fine-tuning transfers across multiple participants.
Research on Optimization of Natural Language Processing Model Based on Multimodal Deep Learning
Sun, Dan, Liang, Yaxin, Yang, Yining, Ma, Yuhan, Zhan, Qishi, Gao, Erdi
This project intends to study the image representation based on attention mechanism and multimodal data. By adding multiple pattern layers to the attribute model, the semantic and hidden layers of image content are integrated. The word vector is quantified by the Word2Vec method and then evaluated by a word embedding convolutional neural network. The published experimental results of the two groups were tested. The experimental results show that this method can convert discrete features into continuous characters, thus reducing the complexity of feature preprocessing. Word2Vec and natural language processing technology are integrated to achieve the goal of direct evaluation of missing image features. The robustness of the image feature evaluation model is improved by using the excellent feature analysis characteristics of a convolutional neural network. This project intends to improve the existing image feature identification methods and eliminate the subjective influence in the evaluation process. The findings from the simulation indicate that the novel approach has developed is viable, effectively augmenting the features within the produced representations.
- North America > United States > California (0.14)
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > New York (0.04)
- Asia > Singapore (0.04)
MUM vs Bert Large Language Models
Natural language processing (NLP) has come a long way in recent years, and the development of advanced AI models like MUM (Multitask Unified Model) is making the field even more exciting. While BERT (Bidirectional Encoder Representations from Transformers) has been a powerful tool in NLP, MUM is expected to surpass it in terms of functionality and versatility. BERT (Bidirectional Encoder Representations from Transformers) is a natural language processing model that was introduced by Google in October 2018. The model was developed by a team of researchers at Google led by Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT was designed to improve the accuracy of NLP tasks such as sentiment analysis, question answering, and language translation.
Diplomacy game AI can negotiate, form alliances, and persuade people
Meta has debuted a new AI capable of besting human opponents in the game of Diplomacy. The game has been seen "as a near-impossible grand challenge" for AI, Meta wrote in a blog post about the AI, called CICERO. Diplomacy is especially difficult, even compared to complex games like chess and go, because it requires a mastery not of hard and fast rules, but of soft skills. Players must know the art of understanding other people's perspectives and needs, wants and wonts; make complex, living plans that can change with human whims; and then persuade other players to work with them and against others. Because it relies on social -- not strategic, logical, or mathematical -- skills, Diplomacy has long been seen as a "near-impossible" challenge for an AI.
Intro to NVIDIA NeMo -Tutorial & Example
NVIDIA NeMo is a toolkit for building new state-of-the-art conversational AI models. NeMo has separate collections for Automatic Speech Recognition (ASR), Natural Language Processing (NLP), and Text-to-Speech (TTS) models. Each collection consists of prebuilt modules that include everything needed to train on your data. Every module can easily be customized, extended, and composed to create new conversational AI model architectures. So let's explain briefly what is the ASR, NLP, and TTS models.